Goto

Collaborating Authors

 semi-supervised image classification


Exploring Probabilistic Models for Semi-supervised Learning

Wang, Jianfeng

arXiv.org Artificial Intelligence

This thesis studies advanced probabilistic models, including both their theoretical foundations and practical applications, for different semi-supervised learning (SSL) tasks. The proposed probabilistic methods are able to improve the safety of AI systems in real applications by providing reliable uncertainty estimates quickly, and at the same time, achieve competitive performance compared to their deterministic counterparts. The experimental results indicate that the methods proposed in the thesis have great value in safety-critical areas, such as the autonomous driving or medical imaging analysis domain, and pave the way for the future discovery of highly effective and efficient probabilistic approaches in the SSL sector.


Energy-Based Models with Applications to Speech and Language Processing

Ou, Zhijian

arXiv.org Artificial Intelligence

Energy-Based Models (EBMs) are an important class of probabilistic models, also known as random fields and undirected graphical models. EBMs are un-normalized and thus radically different from other popular self-normalized probabilistic models such as hidden Markov models (HMMs), autoregressive models, generative adversarial nets (GANs) and variational auto-encoders (VAEs). Over the past years, EBMs have attracted increasing interest not only from the core machine learning community, but also from application domains such as speech, vision, natural language processing (NLP) and so on, due to significant theoretical and algorithmic progress. The sequential nature of speech and language also presents special challenges and needs a different treatment from processing fix-dimensional data (e.g., images). Therefore, the purpose of this monograph is to present a systematic introduction to energy-based models, including both algorithmic progress and applications in speech and language processing. First, the basics of EBMs are introduced, including classic models, recent models parameterized by neural networks, sampling methods, and various learning methods from the classic learning algorithms to the most advanced ones. Then, the application of EBMs in three different scenarios is presented, i.e., for modeling marginal, conditional and joint distributions, respectively. 1) EBMs for sequential data with applications in language modeling, where the main focus is on the marginal distribution of a sequence itself; 2) EBMs for modeling conditional distributions of target sequences given observation sequences, with applications in speech recognition, sequence labeling and text generation; 3) EBMs for modeling joint distributions of both sequences of observations and targets, and their applications in semi-supervised learning and calibrated natural language understanding.


NP-Match: Towards a New Probabilistic Model for Semi-Supervised Learning

Wang, Jianfeng, Hu, Xiaolin, Lukasiewicz, Thomas

arXiv.org Artificial Intelligence

Semi-supervised learning (SSL) has been widely explored in recent years, and it is an effective way of leveraging unlabeled data to reduce the reliance on labeled data. In this work, we adjust neural processes (NPs) to the semi-supervised image classification task, resulting in a new method named NP-Match. NP-Match is suited to this task for two reasons. Firstly, NP-Match implicitly compares data points when making predictions, and as a result, the prediction of each unlabeled data point is affected by the labeled data points that are similar to it, which improves the quality of pseudo-labels. Secondly, NP-Match is able to estimate uncertainty that can be used as a tool for selecting unlabeled samples with reliable pseudo-labels. Compared with uncertainty-based SSL methods implemented with Monte-Carlo (MC) dropout, NP-Match estimates uncertainty with much less computational overhead, which can save time at both the training and the testing phases. We conducted extensive experiments on five public datasets under three semi-supervised image classification settings, namely, the standard semi-supervised image classification, the imbalanced semi-supervised image classification, and the multi-label semi-supervised image classification, and NP-Match outperforms state-of-the-art (SOTA) approaches or achieves competitive results on them, which shows the effectiveness of NP-Match and its potential for SSL. The codes are at https://github.com/Jianf-Wang/NP-Match


Semi-supervised image classification explained

@machinelearnbot

Semi-supervised machine learning is getting ready for primetime. In this article we review a number of common semi-supervised algorithms, capped by a presentation of our own Mean Teacher [arxiv, github], presented at NIPS 2017. Deep learning models have delivered superhuman performance for many years. However, training with standard supervised techniques requires huge amounts of correctly labeled data. Being able to use unlabeled data would open doors to many new applications in e.g.


[R] Semi-supervised image classification explained • r/MachineLearning

@machinelearnbot

I can't seem to find a CIFAR version with pre-deactivated labels for the unlabeled samples. I suppose if you use it as a semi-supervised benchmark you always retain the same labeled records, don't you? I would run the graph based methods I wrote my master thesis on against it for fun, but I'm afraid my personal workstation hasn't got enough ram for 60000x60000 matrices.


Noise-Robust Semi-Supervised Learning by Large-Scale Sparse Coding

Lu, Zhiwu (Renmin University of China) | Gao, Xin (King Abdullah University of Science and Technology) | Wang, Liwei (Peking University) | Wen, Ji-Rong (Renmin University of China) | Huang, Songfang (IBM China Research Lab)

AAAI Conferences

This paper presents a large-scale sparse coding algorithm to deal with the challenging problem of noise-robust semi-supervised learning over very large data with only few noisy initial labels. By giving an L1-norm formulation of Laplacian regularization directly based upon the manifold structure of the data, we transform noise-robust semi-supervised learning into a generalized sparse coding problem so that noise reduction can be imposed upon the noisy initial labels. Furthermore, to keep the scalability of noise-robust semi-supervised learning over very large data, we make use of both nonlinear approximation and dimension reduction techniques to solve this generalized sparse coding problem in linear time and space complexity. Finally, we evaluate the proposed algorithm in the challenging task of large-scale semi-supervised image classification with only few noisy initial labels. The experimental results on several benchmark image datasets show the promising performance of the proposed algorithm.